301 research outputs found
Improving Recommendation Quality by Merging Collaborative Filtering and Social Relationships
Matrix Factorization techniques have been successfully applied to raise the quality of suggestions generated\ud
by Collaborative Filtering Systems (CFSs). Traditional CFSs\ud
based on Matrix Factorization operate on the ratings provided\ud
by users and have been recently extended to incorporate\ud
demographic aspects such as age and gender. In this paper we\ud
propose to merge CF techniques based on Matrix Factorization\ud
and information regarding social friendships in order to\ud
provide users with more accurate suggestions and rankings\ud
on items of their interest. The proposed approach has been\ud
evaluated on a real-life online social network; the experimental\ud
results show an improvement against existing CF approaches.\ud
A detailed comparison with related literature is also presen
XML Matchers: approaches and challenges
Schema Matching, i.e. the process of discovering semantic correspondences
between concepts adopted in different data source schemas, has been a key topic
in Database and Artificial Intelligence research areas for many years. In the
past, it was largely investigated especially for classical database models
(e.g., E/R schemas, relational databases, etc.). However, in the latest years,
the widespread adoption of XML in the most disparate application fields pushed
a growing number of researchers to design XML-specific Schema Matching
approaches, called XML Matchers, aiming at finding semantic matchings between
concepts defined in DTDs and XSDs. XML Matchers do not just take well-known
techniques originally designed for other data models and apply them on
DTDs/XSDs, but they exploit specific XML features (e.g., the hierarchical
structure of a DTD/XSD) to improve the performance of the Schema Matching
process. The design of XML Matchers is currently a well-established research
area. The main goal of this paper is to provide a detailed description and
classification of XML Matchers. We first describe to what extent the
specificities of DTDs/XSDs impact on the Schema Matching task. Then we
introduce a template, called XML Matcher Template, that describes the main
components of an XML Matcher, their role and behavior. We illustrate how each
of these components has been implemented in some popular XML Matchers. We
consider our XML Matcher Template as the baseline for objectively comparing
approaches that, at first glance, might appear as unrelated. The introduction
of this template can be useful in the design of future XML Matchers. Finally,
we analyze commercial tools implementing XML Matchers and introduce two
challenging issues strictly related to this topic, namely XML source clustering
and uncertainty management in XML Matchers.Comment: 34 pages, 8 tables, 7 figure
Measuring Similarity in Large-Scale Folksonomies
Social (or folksonomic) tagging has become a very popular way to describe content within Web 2.0 websites. Unlike\ud
taxonomies, which overimpose a hierarchical categorisation of content, folksonomies enable end-users to freely create and choose the categories (in this case, tags) that best\ud
describe some content. However, as tags are informally de-\ud
fined, continually changing, and ungoverned, social tagging\ud
has often been criticised for lowering, rather than increasing, the efficiency of searching, due to the number of synonyms, homonyms, polysemy, as well as the heterogeneity of\ud
users and the noise they introduce. To address this issue, a\ud
variety of approaches have been proposed that recommend\ud
users what tags to use, both when labelling and when looking for resources. As we illustrate in this paper, real world\ud
folksonomies are characterized by power law distributions\ud
of tags, over which commonly used similarity metrics, including the Jaccard coefficient and the cosine similarity, fail\ud
to compute. We thus propose a novel metric, specifically\ud
developed to capture similarity in large-scale folksonomies,\ud
that is based on a mutual reinforcement principle: that is,\ud
two tags are deemed similar if they have been associated to\ud
similar resources, and vice-versa two resources are deemed\ud
similar if they have been labelled by similar tags. We offer an efficient realisation of this similarity metric, and assess its quality experimentally, by comparing it against cosine similarity, on three large-scale datasets, namely Bibsonomy, MovieLens and CiteULike
Analyzing the Facebook Friendship Graph
Online Social Networks (OSN) during last years acquired a\ud
huge and increasing popularity as one of the most important emerging Web phenomena, deeply modifying the behavior of users and contributing to build a solid substrate of connections and relationships among people using the Web. In this preliminary work paper, our purpose is to analyze Facebook, considering a signi�cant sample of data re\ud
ecting relationships among subscribed users. Our goal is to extract, from this platform, relevant information about the distribution of these relations and exploit tools and algorithms provided by the Social Network Analysis (SNA) to discover and, possibly, understand underlying similarities\ud
between the developing of OSN and real-life social networks
On Facebook, most ties are weak
Pervasive socio-technical networks bring new conceptual and technological
challenges to developers and users alike. A central research theme is
evaluation of the intensity of relations linking users and how they facilitate
communication and the spread of information. These aspects of human
relationships have been studied extensively in the social sciences under the
framework of the "strength of weak ties" theory proposed by Mark Granovetter.13
Some research has considered whether that theory can be extended to online
social networks like Facebook, suggesting interaction data can be used to
predict the strength of ties. The approaches being used require handling
user-generated data that is often not publicly available due to privacy
concerns. Here, we propose an alternative definition of weak and strong ties
that requires knowledge of only the topology of the social network (such as who
is a friend of whom on Facebook), relying on the fact that online social
networks, or OSNs, tend to fragment into communities. We thus suggest
classifying as weak ties those edges linking individuals belonging to different
communities and strong ties as those connecting users in the same community. We
tested this definition on a large network representing part of the Facebook
social graph and studied how weak and strong ties affect the
information-diffusion process. Our findings suggest individuals in OSNs
self-organize to create well-connected communities, while weak ties yield
cohesion and optimize the coverage of information spread.Comment: Accepted version of the manuscript before ACM editorial work. Check
http://cacm.acm.org/magazines/2014/11/179820-on-facebook-most-ties-are-weak/
for the final versio
Enhancing community detection using a network weighting strategy
A community within a network is a group of vertices densely connected to each
other but less connected to the vertices outside. The problem of detecting
communities in large networks plays a key role in a wide range of research
areas, e.g. Computer Science, Biology and Sociology. Most of the existing
algorithms to find communities count on the topological features of the network
and often do not scale well on large, real-life instances.
In this article we propose a strategy to enhance existing community detection
algorithms by adding a pre-processing step in which edges are weighted
according to their centrality w.r.t. the network topology. In our approach, the
centrality of an edge reflects its contribute to making arbitrary graph
tranversals, i.e., spreading messages over the network, as short as possible.
Our strategy is able to effectively complements information about network
topology and it can be used as an additional tool to enhance community
detection. The computation of edge centralities is carried out by performing
multiple random walks of bounded length on the network. Our method makes the
computation of edge centralities feasible also on large-scale networks. It has
been tested in conjunction with three state-of-the-art community detection
algorithms, namely the Louvain method, COPRA and OSLOM. Experimental results
show that our method raises the accuracy of existing algorithms both on
synthetic and real-life datasets.Comment: 28 pages, 2 figure
Effective Retrieval of Resources in Folksonomies Using a New Tag Similarity Measure
Social (or folksonomic) tagging has become a very popular way to describe
content within Web 2.0 websites. However, as tags are informally defined,
continually changing, and ungoverned, it has often been criticised for
lowering, rather than increasing, the efficiency of searching. To address this
issue, a variety of approaches have been proposed that recommend users what
tags to use, both when labeling and when looking for resources. These
techniques work well in dense folksonomies, but they fail to do so when tag
usage exhibits a power law distribution, as it often happens in real-life
folksonomies. To tackle this issue, we propose an approach that induces the
creation of a dense folksonomy, in a fully automatic and transparent way: when
users label resources, an innovative tag similarity metric is deployed, so to
enrich the chosen tag set with related tags already present in the folksonomy.
The proposed metric, which represents the core of our approach, is based on the
mutual reinforcement principle. Our experimental evaluation proves that the
accuracy and coverage of searches guaranteed by our metric are higher than
those achieved by applying classical metrics.Comment: 6 pages, 2 figures, CIKM 2011: 20th ACM Conference on Information and
Knowledge Managemen
- …